assistive technology
AccessEval: Benchmarking Disability Bias in Large Language Models
Panda, Srikant, Agarwal, Amit, Patel, Hitesh Laxmichand
Large Language Models (LLMs) are increasingly deployed across diverse domains but often exhibit disparities in how they handle real-life queries. To systematically investigate these effects within various disability contexts, we introduce \textbf{AccessEval (Accessibility Evaluation)}, a benchmark evaluating 21 closed- and open-source LLMs across 6 real-world domains and 9 disability types using paired Neutral and Disability-Aware Queries. We evaluated model outputs with metrics for sentiment, social perception, and factual accuracy. Our analysis reveals that responses to disability-aware queries tend to have a more negative tone, increased stereotyping, and higher factual error compared to neutral queries. These effects show notable variation by domain and disability type, with disabilities affecting hearing, speech, and mobility disproportionately impacted. These disparities reflect persistent forms of ableism embedded in model behavior. By examining model performance in real-world decision-making contexts, we better illuminate how such biases can translate into tangible harms for disabled users. This framing helps bridges the gap between technical evaluation and user impact, reinforcing importance of bias mitigation in day-to-day applications. Our dataset is publicly available at: https://huggingface.co/datasets/Srikant86/AccessEval
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- North America > United States > Pennsylvania (0.04)
- North America > United States > Georgia > Fulton County > Atlanta (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.94)
- Law (1.00)
- Banking & Finance (1.00)
- Health & Medicine > Therapeutic Area > Psychiatry/Psychology (0.95)
- (4 more...)
Development of a Neural Network Model for Currency Detection to aid visually impaired people in Nigeria
Nwokoye, Sochukwuma, Moru, Desmond
Neural networks in assistive technology for visually impaired leverage artificial intelligence's capacity to recognize patterns in complex data. They are used for converting visual data into auditory or tactile representations, helping the visually impaired understand their surroundings. The primary aim of this research is to explore the potential of artificial neural networks to facilitate the differentiation of various forms of cash for individuals with visual impairments. In this study, we built a custom dataset of 3,468 images, which was subsequently used to train an SSD neural network model. The proposed system can accurately identify Nigerian cash, thereby streamlining commercial transactions. The performance of the system in terms of accuracy was assessed, and the Mean Average Precision score was over 90%. We believe that our system has the potential to make a substantial contribution to the field of assistive technology while also improving the quality of life of visually challenged persons in Nigeria and beyond.
HapticVLM: VLM-Driven Texture Recognition Aimed at Intelligent Haptic Interaction
Khan, Muhammad Haris, Cabrera, Miguel Altamirano, Iarchuk, Dmitrii, Mahmoud, Yara, Trinitatova, Daria, Tokmurziyev, Issatay, Tsetserukou, Dzmitry
-- This paper introduces HapticVLM, a novel mul-timodal system that integrates vision-language reasoning with deep convolutional networks to enable real-time haptic feedback. HapticVLM leverages a ConvNeXt-based material recognition module to generate robust visual embeddings for accurate identification of object materials, while a state-of-the-art Vision-Language Model (Qwen2-VL-2B-Instruct) infers ambient temperature from environmental cues. Experimental evaluations demonstrate an average recognition accuracy of 84.67% across five distinct auditory-tactile patterns and a temperature estimation accuracy of 86.7% based on a tolerance-based evaluation method with an 8 C margin of error across 15 scenarios. Although promising, the current study is limited by the use of a small set of prominent patterns and a modest participant pool. Future work will focus on expanding the range of tactile patterns and increasing user studies to further refine and validate the system's performance. Overall, HapticVLM presents a significant step toward context-aware, multimodal haptic interaction with potential applications in virtual reality, and assistive technologies. I. INTRODUCTION The ability to perceive and distinguish material properties such as texture, temperature, and stiffness is a fundamental aspect of human interaction with the physical world. Human tactile perception integrates visual, auditory, and haptic cues to form a comprehensive understanding of object surfaces, enabling precise material recognition and interaction [1].
- Research Report > New Finding (0.47)
- Research Report > Experimental Study (0.46)
WebNav: An Intelligent Agent for Voice-Controlled Web Navigation
Srinivasan, Trisanth, Patapati, Santosh
The increasing reliance on web interfaces presents many challenges for visually impaired users, showcasing the need for more advanced assistive technologies. This paper introduces WebNav, a voice-controlled web navigation agent that leverages a ReAct-inspired architecture and generative AI to provide this framework. WebNav comprises of a hierarchical structure: a Digital Navigation Module (DIGNAV) for high-level strategic planning, an Assistant Module for translating abstract commands into executable actions, and an Inference Module for low-level interaction. A key component is a dynamic labeling engine, implemented as a browser extension, that generates real-time labels for interactive elements, creating mapping between voice commands and Document Object Model (DOM) components. Preliminary evaluations show that WebNav outperforms traditional screen readers in response time and task completion accuracy for the visually impaired. Future work will focus on extensive user evaluations, benchmark development, and refining the agent's adaptive capabilities for real-world deployment.
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (1.00)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Speech > Speech Recognition (0.89)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.69)
Accessibility Considerations in the Development of an AI Action Plan
Mankoff, Jennifer, Light, Janice, Coughlan, James, Vogler, Christian, Glasser, Abraham, Vanderheiden, Gregg, Rice, Laura
AI has the potential to empower everyone to become more independent and self-sufficient. The increasing use of artificial intelligence (AI)-based technologies in everyday settings creates new opportunities to understand how disabled people might use these technologies [Glazko, 2023]. It also enables the development of new types of assistive technologies as well as new ways for people with disabilities to interact with technology in ways that are both simpler (for those who need things simpler) and more efficient and effective for those who cannot use the traditional interfaces effectively. AI has been rapidly taken up in almost all accessibility communities [Adnin 2024, Alharbi 2024, Jiang 2024, Bennett 2024, Valencia 2023]. Since becoming widely available to the public, Generative Artificial Intelligence (GAI) has steadily gained recognition for its potential as a valuable tool in the private sector and by government, as well as a tool for accessibility. Studies of blind and visually impaired individuals have found that they use GAI to'offload' cognitively demanding tasks and obtain personal help such as fashion advice (e.g., [Xie 2024]), and to create content or retrieve information [Adnin 2024]. A study of GAI use by neurodiverse users found GAI can both support and complicate tasks like code-switching, emotional regulation, and accessing information [Glazko, 2025]. A study of people who use AAC found it helpful for text input [Valencia 2023]. However there are concerns with a technology that is often based on probability and thus tends toward the most common case rather than those at the margins.
- Information Technology > Security & Privacy (1.00)
- Health & Medicine > Therapeutic Area (0.66)
- Information Technology > Artificial Intelligence > Natural Language > Generation (0.48)
- Information Technology > Artificial Intelligence > Issues > Social & Ethical Issues (0.47)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.34)
VocalEyes: Enhancing Environmental Perception for the Visually Impaired through Vision-Language Models and Distance-Aware Object Detection
Chavan, Kunal, Balaji, Keertan, Barigidad, Spoorti, Chiluveru, Samba Raju
With an increasing demand for assistive technologies that promote the independence and mobility of visually impaired people, this study suggests an innovative real-time system that gives audio descriptions of a user's surroundings to improve situational awareness. The system acquires live video input and processes it with a quantized and fine-tuned Florence-2 big model, adjusted to 4-bit accuracy for efficient operation on low-power edge devices such as the NVIDIA Jetson Orin Nano. By transforming the video signal into frames with a 5-frame latency, the model provides rapid and contextually pertinent descriptions of objects, pedestrians, and barriers, together with their estimated distances. The system employs Parler TTS Mini, a lightweight and adaptable Text-to-Speech (TTS) solution, for efficient audio feedback. It accommodates 34 distinct speaker types and enables customization of speech tone, pace, and style to suit user requirements. This study examines the quantization and fine-tuning techniques utilized to modify the Florence-2 model for this application, illustrating how the integration of a compact model architecture with a versatile TTS component improves real-time performance and user experience. The proposed system is assessed based on its accuracy, efficiency, and usefulness, providing a viable option to aid vision-impaired users in navigating their surroundings securely and successfully.
- Research Report > Experimental Study (0.54)
- Research Report > New Finding (0.47)
- Health & Medicine (0.90)
- Information Technology > Hardware (0.35)
Adaptive Object Detection for Indoor Navigation Assistance: A Performance Evaluation of Real-Time Algorithms
Pratap, Abhinav, Kumar, Sushant, Chakravarty, Suchinton
-- This study addresses the critical need for accurate and efficient object detection in assistive technologies for visually impaired individuals. We systematically evaluate the performance of four prominent real-time object detection algorithms--YOLO, SSD, Faster R-CNN, and Mask R-CNN--within the context of indoor navigation assistance. Our analysis, conducted on the Indoor Objects Detection dataset, focuses on key parameters including detection accuracy, processing speed, and adaptability to the unique challenges of indoor environments. This research contributes to a deeper understanding of adaptive machine learning applications that can significantly improve indoor navigation solutions for the visually impaired, promoting inclusivity and accessibility. In today's technology-driven society, there is an increasing emphasis on enhancing accessibility for visually impaired individuals.
- North America > United States > Minnesota > Blue Earth County > Mankato (0.04)
- Asia > India > Tamil Nadu > Chennai (0.04)
- Asia > India > Maharashtra (0.04)
- Asia > China (0.04)
Small Object Detection for Indoor Assistance to the Blind using YOLO NAS Small and Super Gradients
BN, Rashmi, Guru, R., A, Anusuya M
Advancements in object detection algorithms have opened new avenues for assistive technologies that cater to the needs of visually impaired individuals. This paper presents a novel approach for indoor assistance to the blind by addressing the challenge of small object detection. We propose a technique YOLO NAS Small architecture, a lightweight and efficient object detection model, optimized using the Super Gradients training framework. This combination enables real-time detection of small objects crucial for assisting the blind in navigating indoor environments, such as furniture, appliances, and household items. Proposed method emphasizes low latency and high accuracy, enabling timely and informative voice-based guidance to enhance the user's spatial awareness and interaction with their surroundings. The paper details the implementation, experimental results, and discusses the system's effectiveness in providing a practical solution for indoor assistance to the visually impaired.
Evaluating Assistive Technologies on a Trade Fair: Methodological Overview and Lessons Learned
Baumeister, Annalies, Goldau, Felix, Pascher, Max, Gerken, Jens, Frese, Udo, Tolle, Patrizia
User-centered evaluations are a core requirement in the development of new user related technologies. However, it is often difficult to recruit sufficient participants, especially if the target population is small, particularly busy, or in some way restricted in their mobility. We bypassed these problems by conducting studies on trade fairs that were specifically designed for our target population (potentially care-receiving individuals in wheelchairs) and therefore provided our users with external incentive to attend our study. This paper presents our gathered experiences, including methodological specifications and lessons learned, and is aimed to guide other researchers with conducting similar studies. In addition, we also discuss chances generated by this unconventional study environment as well as its limitations.
- North America > Canada > Ontario (0.14)
- North America > United States > New York > New York County > New York City (0.05)
- Europe > Germany > North Rhine-Westphalia > Düsseldorf Region > Düsseldorf (0.04)
- (15 more...)
- Research Report > New Finding (0.68)
- Research Report > Experimental Study (0.46)
- Health & Medicine > Therapeutic Area > Neurology (0.93)
- Information Technology > Security & Privacy (0.68)
ProgramAlly: Creating Custom Visual Access Programs via Multi-Modal End-User Programming
Herskovitz, Jaylin, Xu, Andi, Alharbi, Rahaf, Guo, Anhong
Existing visual assistive technologies are built for simple and common use cases, and have few avenues for blind people to customize their functionalities. Drawing from prior work on DIY assistive technology, this paper investigates end-user programming as a means for users to create and customize visual access programs to meet their unique needs. We introduce ProgramAlly, a system for creating custom filters for visual information, e.g., 'find NUMBER on BUS', leveraging three end-user programming approaches: block programming, natural language, and programming by example. To implement ProgramAlly, we designed a representation of visual filtering tasks based on scenarios encountered by blind people, and integrated a set of on-device and cloud models for generating and running these programs. In user studies with 12 blind adults, we found that participants preferred different programming modalities depending on the task, and envisioned using visual access programs to address unique accessibility challenges that are otherwise difficult with existing applications. Through ProgramAlly, we present an exploration of how blind end-users can create visual access programs to customize and control their experiences.
- North America > United States > Michigan > Washtenaw County > Ann Arbor (0.14)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.05)
- (3 more...)
- Health & Medicine (0.93)
- Education (0.92)
- Information Technology (0.67)